# Efficient Pre-training
Open Qwen2VL
CC
Open-Qwen2VL is a multimodal model capable of receiving both images and text as input and generating text output.
Image-to-Text English
O
weizhiwang
568
15
Llama3 German 8B 32k
A German-optimized large language model based on Meta Llama3-8B, continuously pre-trained on 65 billion German tokens, specifically optimized for German and supporting 32k long context
Large Language Model
Transformers German

L
DiscoResearch
91
13
Tinyllama V1.1
Apache-2.0
TinyLlama is a small language model with 1.1 billion parameters, adopting the same architecture and tokenizer as Llama 2, suitable for resource-constrained application scenarios.
Large Language Model
Transformers English

T
TinyLlama
42.11k
92
Videomae Base
VideoMAE is a video self-supervised pretraining model based on Masked Autoencoder (MAE), which learns internal video representations by predicting pixel values of masked video patches.
Video Processing
Transformers

V
MCG-NJU
48.66k
45
Chinese Electra Large Generator
Apache-2.0
Chinese ELECTRA is a pre-trained model developed by the HIT-iFLYTEK Joint Lab based on Google's ELECTRA model, featuring a small parameter size but superior performance.
Large Language Model
Transformers Chinese

C
hfl
14
0
Distilcamembert Base
MIT
DistilCamemBERT is a distilled version of the French CamemBERT model, significantly reducing model complexity while maintaining performance through knowledge distillation techniques.
Large Language Model
Transformers French

D
cmarkea
15.79k
31
Chinese Mobile Bert
Apache-2.0
This model was pre-trained on a 250-million-word Chinese corpus using the MobileBERT architecture, with a training period of 15 days, completing 1 million iterations on a single A100 GPU.
Large Language Model
Transformers

C
Ayou
25
5
Featured Recommended AI Models